historical contrastive learning
Model Adaptation: Historical Contrastive Learning for Unsupervised Domain Adaptation without Source Data
Unsupervised domain adaptation aims to align a labeled source domain and an unlabeled target domain, but it requires to access the source data which often raises concerns in data privacy, data portability and data transmission efficiency. We study unsupervised model adaptation (UMA), or called Unsupervised Domain Adaptation without Source Data, an alternative setting that aims to adapt source-trained models towards target distributions without accessing source data. To this end, we design an innovative historical contrastive learning (HCL) technique that exploits historical source hypothesis to make up for the absence of source data in UMA. HCL addresses the UMA challenge from two perspectives. First, it introduces historical contrastive instance discrimination (HCID) that learns from target samples by contrasting their embeddings which are generated by the currently adapted model and the historical models. With the historical models, HCID encourages UMA to learn instance-discriminative target representations while preserving the source hypothesis. Second, it introduces historical contrastive category discrimination (HCCD) that pseudo-labels target samples to learn category-discriminative target representations.
Model Adaptation: Historical Contrastive Learning for Unsupervised Domain Adaptation without Source Data Supplemental Materials Anonymous Author(s) Affiliation Address email
Maximum likelihood (ML) is a concept to describe the theoretic insights of clustering algorithms. As it is not easy to optimize Eq.14 directly, we employ a surrogate function to lower-bound the Please note the notation " t m" shows that the k is encoded by a historical encoder. In practice, we achieve Eq. 9 by minimizing a historical contrastive instance discrimination loss: Please note that Eq. 10 is an instance of Eq. 9. The two equations look different due to: 1) Eq. 10 Proposition 2 The HCID is convergent under certain conditions. We have illustrated in Section A.1 that the inequality in Eq.11 holds with equality if One possible way to achieve Eq. 13 is to conduct gradient descent by minimizing the historical Different from the classical expectation maximization (mentioned in Section A.1) that consists It can be observed that Eq. 16 is the same as Eq. 15 except involving an extra weighting element In the following, we show the optimization of Eq. 16 is a CEM process.
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.70)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Uncertainty > Bayesian Inference (0.36)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Directed Networks > Bayesian Learning (0.36)
Model Adaptation: Historical Contrastive Learning for Unsupervised Domain Adaptation without Source Data
Unsupervised domain adaptation aims to align a labeled source domain and an unlabeled target domain, but it requires to access the source data which often raises concerns in data privacy, data portability and data transmission efficiency. We study unsupervised model adaptation (UMA), or called Unsupervised Domain Adaptation without Source Data, an alternative setting that aims to adapt source-trained models towards target distributions without accessing source data. To this end, we design an innovative historical contrastive learning (HCL) technique that exploits historical source hypothesis to make up for the absence of source data in UMA. HCL addresses the UMA challenge from two perspectives. First, it introduces historical contrastive instance discrimination (HCID) that learns from target samples by contrasting their embeddings which are generated by the currently adapted model and the historical models.
Temporal Knowledge Graph Reasoning with Historical Contrastive Learning
Xu, Yi, Ou, Junjie, Xu, Hui, Fu, Luoyi
Temporal knowledge graph, serving as an effective way to store and model dynamic relations, shows promising prospects in event forecasting. However, most temporal knowledge graph reasoning methods are highly dependent on the recurrence or periodicity of events, which brings challenges to inferring future events related to entities that lack historical interaction. In fact, the current moment is often the combined effect of a small part of historical information and those unobserved underlying factors. To this end, we propose a new event forecasting model called Contrastive Event Network (CENET), based on a novel training framework of historical contrastive learning. CENET learns both the historical and non-historical dependency to distinguish the most potential entities that can best match the given query. Simultaneously, it trains representations of queries to investigate whether the current moment depends more on historical or non-historical events by launching contrastive learning. The representations further help train a binary classifier whose output is a boolean mask to indicate related entities in the search space. During the inference process, CENET employs a mask-based strategy to generate the final results. We evaluate our proposed model on five benchmark graphs. The results demonstrate that CENET significantly outperforms all existing methods in most metrics, achieving at least $8.3\%$ relative improvement of Hits@1 over previous state-of-the-art baselines on event-based datasets.
- Asia > North Korea (0.05)
- North America > United States (0.04)
- Asia > China > Shanghai > Shanghai (0.04)
- (4 more...)